Avoiding Bad Query Mixes to Minimize Unsuccessful Client Requests Under Heavy Loads
نویسنده
چکیده
In three-tiered web applications, some form of admission control is required to ensure that throughput and response times are not significantly harmed during periods of heavy load. We propose Q-Cop, a prototype system for improving admission control decisions that computes measures of load on the system based on the actual mix of queries being executed. This measure of load is used to estimate execution times for incoming queries, which allows Q-Cop to make control decisions with the goal of minimizing the number of requests that are not serviced before the client, or their browser, times out. Using TPC-W queries, we show that the response times of different types of queries can vary significantly, in excess of 50% in our experiments, depending not just on the number of queries being processed but on the mix of other queries that are running simultaneously. The variation implies that admission control can benefit from taking into account not just the number of queries being processed, but also the mix of queries. We develop a model of expected query execution times that accounts for the mix of queries being executed and integrate that model into a three-tiered system to make admission control decisions. This approach makes more informed decisions about which queries to reject, and our results show that it significantly reduces the number of unsuccessful client requests. Our results show that this approach makes more informed decisions about which queries to reject and as a result significantly reduces the number of unsuccessful client requests. For comparison, we develop several other models which represent related work in the field, including an MPL-based approach and an approach that considers the type of query but not the mix of queries. We show that Q-Cop does not need to re-compute any modelling information in order to perform well, a strong advantage over most other approaches. Across the range of workloads examined, an average of 47% fewer requests are denied than the next best approach.
منابع مشابه
Enhancing GridFTP performance using intelligent gateways
To improve throughput of grid data servers under heavy loads or under denial of service attacks, it is important to service requests differentially, giving preference to ongoing or imminent client requests. We show how such features can be efficiently implemented on a gateway that controls access to a pool of servers. We present performance results for a prototype system based on a dual-ported ...
متن کاملDetecting Timed-Out Client Requests for Avoiding Livelock and Improving Web Server Performance
A Web server’s listen queue capacity is often set to a large value to accommodate bursts of traffic and to accept as many client requests as possible. We show that, under certain overload conditions, this results in a significant loss of server performance due to the processing of so-called “dead requests”: timed-out client requests whose associated connection has been closed from the client si...
متن کاملMulticasting a Changing Repository
Web crawlers generate significant loads on Web servers, and are difficult to operate. Instead of repeatedly running crawlers at many “client” sites, we propose a central crawler and Web repository that multicasts appropriate subsets of the central repository, and their subsequent changes, to subscribing clients. Loads at Web servers are reduced because a single crawler visits the servers, as op...
متن کاملSilo: Exploiting JavaScript and DOM Storage for Faster Page Loads
A modern web page contains many objects, and fetching these objects requires many network round trips— establishing each HTTP connection requires a TCP handshake, and each HTTP request/response pair requires at least one round trip. To decrease a page’s load time, designers try to minimize the number of HTTP requests needed to fetch the constituent objects. A common strategy is to inline the pa...
متن کاملSWIFT: Scheduling in Web Servers for Fast Response Time
This paper addresses the problem of how to service web requests quickly in order to minimize the client response time. Some of the recent work uses the idea of the Shortest Remaining Processing Time scheduling (SRPT) in Web servers in order to give preference to requests for short files. However, by considering only the size of the file for determining the priority of requests, the previous wor...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
عنوان ژورنال:
دوره شماره
صفحات -
تاریخ انتشار 2009